Learning from MOM’s principles : Le Cam’s approach
نویسندگان
چکیده
We obtain estimation error rates for estimators obtained by aggregation of regularized median-of-means tests, following a construction of Le Cam. The results hold with exponentially large probability, under only weak moments assumptions on data. Any norm may be used for regularization. When it has some sparsity inducing power we recover sparse rates of convergence. The procedure is robust since a large part of data may be corrupted, these outliers have nothing to do with the oracle we want to reconstruct. Our general risk bound is of order max ( minimax rate in the i.i.d. setup, number of outliers number of observations ) . In particular, the number of outliers may be as large as (number of data) ×(minimax rate) without affecting this rate. The other data do not have to be identically distributed but should only have equivalent L and L moments. For example, the minimax rate s log(ed/s)/N of recovery of a s-sparse vector in R is achieved with exponentially large probability by a median-of-means version of the LASSO when the noise has q0 moments for some q0 > 2, the entries of the design matrix should have C0 log(ed) moments and the dataset can be corrupted up to C1s log(ed/s) outliers.
منابع مشابه
A New Approach to Poisson Approximations∗
The main purpose of this note is to present a new approach to Poisson Approximations. Some bounds in Poisson Approximations in term of classical Le Cam’s inequalities for various row-wise triangular arrays of independent Poisson-binomial distributed random variables are established via probability distances based on Trotter-Renyi operators. Some analogous results related to random sums in Poiss...
متن کاملLecture 21: Minimax Theory
We also saw how to use the Neyman-Pearson lemma to derive a lower bound when M = 2, which is the simple versus simple hypothesis testing case. This is known as Le Cam’s method At the end of last class, I mentioned how Le Cam’s method doesn’t work well for multi-dimensional estimation problems. To see why, let’s revisit the gaussian mean estimation problem where ρ = ‖ · ‖2 and Φ(t) = t, but now ...
متن کاملAsymptotic equivalence for nonparametric regression with multivariate and random design
We show that nonparametric regression is asymptotically equivalent in Le Cam’s sense with a sequence of Gaussian white noise experiments as the number of observations tends to infinity. We propose a general constructive framework based on approximation spaces, which permits to achieve asymptotic equivalence even in the cases of multivariate and random design.
متن کاملAsymptotic Equivalence of Density Estimation and Gaussian White Noise
Signal recovery in Gaussian white noise with variance tending to zero has served for some time as a representative model for nonparametric curve estimation, having all the essential traits in a pure form. The equivalence has mostly been stated informally, but an approximation in the sense of Le Cam’s deficiency distance 1 would make it precise. The models are then asymptotically equivalent for ...
متن کاملChapter 3 Assouad ’ s method
Assouad’s method provides a somewhat different technique for proving lower bounds. Instead of reducing the estimation problem to a multiple hypothesis test or simpler estimation problem, as with Le Cam’s method and Fano’s method from the preceding lectures, here we transform the original estimation problem into multiple binary hypothesis testing problems, using the structure of the problem in a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017